Objective: We aim to develop an open-source natural language processing (NLP) package, SODA (i.e., SOcial DeterminAnts), with pre-trained transformer models to extract social determinants of health (SDoH) for cancer patients, examine the generalizability of SODA to a new disease domain (i.e., opioid use), and evaluate the extraction rate of SDoH using cancer populations. Methods: We identified SDoH categories and attributes and developed an SDoH corpus using clinical notes from a general cancer cohort. We compared four transformer-based NLP models to extract SDoH, examined the generalizability of NLP models to a cohort of patients prescribed with opioids, and explored customization strategies to improve performance. We applied the best NLP model to extract 19 categories of SDoH from the breast (n=7,971), lung (n=11,804), and colorectal cancer (n=6,240) cohorts. Results and Conclusion: We developed a corpus of 629 cancer patients notes with annotations of 13,193 SDoH concepts/attributes from 19 categories of SDoH. The Bidirectional Encoder Representations from Transformers (BERT) model achieved the best strict/lenient F1 scores of 0.9216 and 0.9441 for SDoH concept extraction, 0.9617 and 0.9626 for linking attributes to SDoH concepts. Fine-tuning the NLP models using new annotations from opioid use patients improved the strict/lenient F1 scores from 0.8172/0.8502 to 0.8312/0.8679. The extraction rates among 19 categories of SDoH varied greatly, where 10 SDoH could be extracted from >70% of cancer patients, but 9 SDoH had a low extraction rate (<70% of cancer patients). The SODA package with pre-trained transformer models is publicly available at https://github.com/uf-hobiinformatics-lab/SDoH_SODA.
translated by 谷歌翻译
昼夜节律的破坏是阿尔茨海默氏病(AD)患者的基本症状。人类脑中基因表达的完整昼夜节律编排及其与AD的固有关联仍然很大程度上是未知的。我们提出了一种新颖的综合方法,即Prime,以检测和分析在多个数据集中不合时宜的高维基因表达数据中的节奏振荡模式。为了证明Prime的实用性,首先,我们通过从小鼠肝脏中的时间课程表达数据集作为跨物种和跨器官验证来对其进行验证。然后,我们将其应用于研究来自19个对照和AD患者的19个人脑区域的未接收基因组基因表达中的振荡模式。我们的发现揭示了15对控制大脑区域中清晰,同步的振荡模式,而这些振荡模式要么消失或昏暗。值得注意的是,Prime在不需要样品的时间戳而发现昼夜节律的节奏模式。 Prime的代码以及在本文中复制数字的代码,可在https://github.com/xinxingwu-uk/prime上获得。
translated by 谷歌翻译
在本文中,我们描述了一种数据驱动的方法,用于开发艾米丽(Emily),一种情绪感染的开放域聊天机器人。提出的数据增强方法可以从多转话对话中明确模拟阳性过渡(PT)情感数据。我们使用PT情感数据构建对话语料库,并将其发布供公众使用。通过使用生产的PT增强对话进行验证的对话模型,我们能够开发一种情感感染性的开放式聊天机器人,该聊天机器人在各种情绪影响度指标中表现出几乎人类的表现。我们对艾米丽(Emily)进行评估,以针对一些最先进的(SOTA)开放域聊天机器人,并显示拟议方法的有效性。
translated by 谷歌翻译
现有的最先进的(SOTA)视频显着对象检测(VSOD)模型已广泛遵循短期方法,该方法通过仅考虑当前连续的有限帧而动态地确定空间和时间显着性融合之间的平衡。但是,短期方法论具有一个关键限制,这与我们视觉系统的真实机制相抵触,这是一种典型的长期方法。结果,故障案例不断出现在当前的SOTA模型的结果中,而短期方法论成为主要的技术瓶颈。为了解决这个问题,本文提出了一种新颖的VSOD方法,该方法以完整的长期方式执行了VSOD。我们的方法将顺序vSOD(一个顺序任务)转换为数据挖掘问题,即将输入视频序列分解为对象提案,然后尽可能易于挖掘出明显的对象建议。由于所有对象提案都可以同时获得,因此提出的方法是一种完整的长期方法,可以减轻植根于常规短期方法的一些困难。此外,我们设计了一个在线更新方案,该方案可以掌握显着对象的最具代表性和可信赖的模式概况,并使用丰富的细节输出框架显着图,并在空间和时间上平滑。所提出的方法在五个广泛使用的基准数据集上几乎优于所有SOTA模型。
translated by 谷歌翻译
在本文中,我们提出了一种用于HSI去噪的强大主成分分析的新型非耦合方法,其侧重于分别同时为低级和稀疏组分的等级和列方向稀疏性产生更准确的近似。特别是,新方法采用日志确定级别近似和新颖的$ \ ell_ {2,\ log} $常规,以便分别限制组件矩阵的本地低级或列明智地稀疏属性。对于$ \ ell_ {2,\ log} $ - 正常化的收缩问题,我们开发了一个高效的封闭式解决方案,该解决方案名为$ \ ell_ {2,\ log} $ - 收缩运算符。新的正则化和相应的操作员通常可以用于需要列明显稀疏性的其他问题。此外,我们在基于日志的非凸rpca模型中强加了空间光谱总变化正则化,这增强了从恢复的HSI中的空间和光谱视图中的全局转换平滑度和光谱一致性。关于模拟和实际HSIS的广泛实验证明了所提出的方法在去噪HSIS中的有效性。
translated by 谷歌翻译
文本分类是自然语言处理(NLP)的主要任务。最近,图神经网络(GNN)已迅速发展,并应用于文本分类任务。作为一种特殊的图形数据,该树具有更简单的数据结构,可以为文本分类提供丰富的层次结构信息。受结构熵的启发,我们通过最小化结构熵并提出提示来构造图形的编码树,该提示旨在充分利用文本中包含的文本中包含的层次信息,以完成文本分类的任务。具体来说,我们首先为每个文本建立依赖关系解析图。然后,我们设计了一种结构熵最小化算法来解码图中的关键信息,并将每个图转换为其相应的编码树。基于编码树的层次结构,通过逐层更新编码树中的非叶子节点的表示来获得整个图的表示。最后,我们介绍了层次信息在文本分类中的有效性。实验结果表明,在具有简单的结构和很少的参数的同时,提示在流行基准测试上的最新方法优于最先进的方法。
translated by 谷歌翻译
In recent years, arbitrary image style transfer has attracted more and more attention. Given a pair of content and style images, a stylized one is hoped that retains the content from the former while catching style patterns from the latter. However, it is difficult to simultaneously keep well the trade-off between the content details and the style features. To stylize the image with sufficient style patterns, the content details may be damaged and sometimes the objects of images can not be distinguished clearly. For this reason, we present a new transformer-based method named STT for image style transfer and an edge loss which can enhance the content details apparently to avoid generating blurred results for excessive rendering on style features. Qualitative and quantitative experiments demonstrate that STT achieves comparable performance to state-of-the-art image style transfer methods while alleviating the content leak problem.
translated by 谷歌翻译
In recent years, the Transformer architecture has shown its superiority in the video-based person re-identification task. Inspired by video representation learning, these methods mainly focus on designing modules to extract informative spatial and temporal features. However, they are still limited in extracting local attributes and global identity information, which are critical for the person re-identification task. In this paper, we propose a novel Multi-Stage Spatial-Temporal Aggregation Transformer (MSTAT) with two novel designed proxy embedding modules to address the above issue. Specifically, MSTAT consists of three stages to encode the attribute-associated, the identity-associated, and the attribute-identity-associated information from the video clips, respectively, achieving the holistic perception of the input person. We combine the outputs of all the stages for the final identification. In practice, to save the computational cost, the Spatial-Temporal Aggregation (STA) modules are first adopted in each stage to conduct the self-attention operations along the spatial and temporal dimensions separately. We further introduce the Attribute-Aware and Identity-Aware Proxy embedding modules (AAP and IAP) to extract the informative and discriminative feature representations at different stages. All of them are realized by employing newly designed self-attention operations with specific meanings. Moreover, temporal patch shuffling is also introduced to further improve the robustness of the model. Extensive experimental results demonstrate the effectiveness of the proposed modules in extracting the informative and discriminative information from the videos, and illustrate the MSTAT can achieve state-of-the-art accuracies on various standard benchmarks.
translated by 谷歌翻译
Learning with noisy-labels has become an important research topic in computer vision where state-of-the-art (SOTA) methods explore: 1) prediction disagreement with co-teaching strategy that updates two models when they disagree on the prediction of training samples; and 2) sample selection to divide the training set into clean and noisy sets based on small training loss. However, the quick convergence of co-teaching models to select the same clean subsets combined with relatively fast overfitting of noisy labels may induce the wrong selection of noisy label samples as clean, leading to an inevitable confirmation bias that damages accuracy. In this paper, we introduce our noisy-label learning approach, called Asymmetric Co-teaching (AsyCo), which introduces novel prediction disagreement that produces more consistent divergent results of the co-teaching models, and a new sample selection approach that does not require small-loss assumption to enable a better robustness to confirmation bias than previous methods. More specifically, the new prediction disagreement is achieved with the use of different training strategies, where one model is trained with multi-class learning and the other with multi-label learning. Also, the new sample selection is based on multi-view consensus, which uses the label views from training labels and model predictions to divide the training set into clean and noisy for training the multi-class model and to re-label the training samples with multiple top-ranked labels for training the multi-label model. Extensive experiments on synthetic and real-world noisy-label datasets show that AsyCo improves over current SOTA methods.
translated by 谷歌翻译
Machine learning models are typically evaluated by computing similarity with reference annotations and trained by maximizing similarity with such. Especially in the bio-medical domain, annotations are subjective and suffer from low inter- and intra-rater reliability. Since annotations only reflect the annotation entity's interpretation of the real world, this can lead to sub-optimal predictions even though the model achieves high similarity scores. Here, the theoretical concept of Peak Ground Truth (PGT) is introduced. PGT marks the point beyond which an increase in similarity with the reference annotation stops translating to better Real World Model Performance (RWMP). Additionally, a quantitative technique to approximate PGT by computing inter- and intra-rater reliability is proposed. Finally, three categories of PGT-aware strategies to evaluate and improve model performance are reviewed.
translated by 谷歌翻译